Near-Minimax Optimal Estimation With Shallow ReLU Neural Networks

نویسندگان

چکیده

We study the problem of estimating an unknown function from noisy data using shallow ReLU neural networks. The estimators we minimize sum squared data-fitting errors plus a regularization term proportional to Euclidean norm network weights. This minimization corresponds common approach training with weight decay. quantify performance (mean-squared error) these when data-generating belongs second-order Radon-domain bounded variation space. space functions was recently proposed as natural associated derive minimax lower bound for estimation this and show that are optimal up logarithmic factors. rate is immune curse dimensionality. explicit gap between networks linear methods (which include kernel methods) by deriving problem, showing necessarily suffer dimensionality in As result, paper sheds light on phenomenon seem break

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Convergence Analysis of Two-layer Neural Networks with ReLU Activation

In recent years, stochastic gradient descent (SGD) based techniques has become the standard tools for training neural networks. However, formal theoretical understanding of why SGD can train neural networks in practice is largely missing. In this paper, we make progress on understanding this mystery by providing a convergence analysis for SGD on a rich subset of two-layer feedforward networks w...

متن کامل

Evolutionary Programming of Near-Optimal Neural Networks

A genetic algorithm (GA) method that evolves both the topology and training parameters of backpropagation-trained, fully-connected, feed-forward neural networks is presented. The GA uses a weak encoding scheme with real-valued alleles. One contribution of the proposed approach is to replace the needed but potentially slow evolution of final weights by the more efficient evolution of a single we...

متن کامل

Automatically searching near-optimal artificial neural networks

The idea of automatically searching neural networks that learn faster and generalize better is becoming increasingly widespread. In this paper, we present a new method for searching near-optimal artificial neural networks that include initial weights, transfer functions, architectures and learning rules that are specially tailored to a given problem. Experimental results have shown that the met...

متن کامل

Nonparametric regression using deep neural networks with ReLU activation function

Consider the multivariate nonparametric regression model. It is shown that estimators based on sparsely connected deep neural networks with ReLU activation function and properly chosen network architecture achieve the minimax rates of convergence (up to log n-factors) under a general composition assumption on the regression function. The framework includes many well-studied structural constrain...

متن کامل

Path-Normalized Optimization of Recurrent Neural Networks with ReLU Activations

We investigate the parameter-space geometry of recurrent neural networks (RNNs), and develop an adaptation of path-SGD optimization method, attuned to this geometry, that can learn plain RNNs with ReLU activations. On several datasets that require capturing long-term dependency structure, we show that path-SGD can significantly improve trainability of ReLU RNNs compared to RNNs trained with SGD...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Transactions on Information Theory

سال: 2023

ISSN: ['0018-9448', '1557-9654']

DOI: https://doi.org/10.1109/tit.2022.3208653